All Questions
743 questions
-1votes
0answers
72views
How to scrape the full New York Times article content using Selenium and BeautifulSoup without triggering the "Please enable JavaScript" message?
I'm building a scraper that fetches full article content from the New York Times using both the Article Search API and a hybrid static + Selenium-based HTML scraper. My goal is to extract complete ...
0votes
1answer
100views
Why can't I extract listings information
I am trying to extract the EPC rating from each listings. you can only get the EPC rating when you click on the listing. each time i run my script it it's keeps timing out, what could be the issue ? ...
0votes
1answer
142views
Can't use captcha solver api services due to missing the element "data-sitekey"
I'm tryingto to use a captcha solving service but all of them request a code that's attached to an element called "data-sitekey" that should be in the HTML of a page with recaptcha on it. ...
0votes
1answer
73views
Get data hidden in ellipses while web scraping
I'm attempting to grab episode title shown at the header of this website. When inspecting the page elements myself I can see near the top a line of HTML like this: <h1 id="epName">...&...
0votes
2answers
77views
How to parse HTML hidden behind JS scripts
The FCC has a database with details about various broadcast licenses. Many of these licenses have pages like this one Most of the data on these pages (and related ones) can be scraped very easily with ...
0votes
1answer
151views
scrape a dropdown list using playwright
I'm struggling to find a way to click on the "All" option in a dropdown list and scrape all the content inside that page. I have come across a few posts but they're a little different from ...
0votes
0answers
84views
scrape link for jwplayer calculated with JS using python
I'm trying to scrape video link (m3u8) from this website: https://deaddrive.xyz/embed/fa31e While inspecting the page, I realized that the link is calculated on the fly using JS in the function: &...
0votes
2answers
182views
How to click the next link with Zyte browser automation?
The Zyte tutorial "Create your first spider" crawls this page which has a pager with a "normal" next link. But what if the next link contains only a href="#" and executes ...
0votes
1answer
38views
How to exclude div classes 'modal-content' and 'modal-body' from pyppeteer web scraper?
I'm building a scraper that gets text data from a list of articles. A common specimen in the text content I'm scraping at the minute is that at the bottom there is this message: "As a subscriber, ...
0votes
1answer
52views
Extracting the text between span tags in a Javascript-rendered page using Selenium in Python
I am trying to scrape all instances of text between tags with a particular class on a web page that dynamically updates. I am using selenium with a chrome WebDriver in Python. In a normal browser, ...
2votes
3answers
95views
scraping table from web page
I'm trying to scrape a table from a webpage using Selenium and BeautifulSoup but I'm not sure how to get to the actual data using BeautifulSoup. webpage: https://leetify.com/app/match-details/5c438e85-...
0votes
1answer
241views
Using Python with Selenium and BeautifulSoup4 how can i get data after Javascript has loaded all elements on the page?
I'm trying to scrape data from a sandbox website just to practice and start using python to scrape web data. I have managed to extract a lot of data using the basics however I have found an element ...
0votes
0answers
65views
Weird API response with <script> JavaScript tag
I am working on a project in Python that scrapes a university portal website to retrieve a weekly schedule. I see from the developer tools that the schedule page makes a API call and receives a JSON ...
0votes
1answer
159views
make client send http request for backend flask
I am trying to avoid rate limiting and ip blacklisting while accessing an external api. I want to deploy a flask web app on google app engine. I need a way to have the client send http requests to the ...
0votes
0answers
23views
Is there a way to mimic the Element.closest() function from javascript in Scrapy python?
I am trying to convert my web-scraper I built in JavaScript using the puppeteer library into a python-based web-scraper running on Scrapy. I want to be able to do something similar to JavaScript's ...